113 research outputs found

    Conversational Functions of Korean Discourse Connective Kulaykaciko

    Get PDF
    The aim of this paper is to examine the use of Korean discourse connective kulaykaciko, one of the discourse connectives frequently produced by participants in Korean conversation. By employing the methodological framework of discourse modality (Maynard, 1993), it investigates semantic, pragmatic and interpersonal functions of kulaykaciko in naturally occurring discourse. The analysis reveals that a single linguistic sign, kulaykaciko has multiple functions in discourse: it can express the cause-result relationship, provide further explanation related to the previous talk, manage turn-taking, and index politeness

    Scale-invariant Bayesian Neural Networks with Connectivity Tangent Kernel

    Full text link
    Explaining generalizations and preventing over-confident predictions are central goals of studies on the loss landscape of neural networks. Flatness, defined as loss invariability on perturbations of a pre-trained solution, is widely accepted as a predictor of generalization in this context. However, the problem that flatness and generalization bounds can be changed arbitrarily according to the scale of a parameter was pointed out, and previous studies partially solved the problem with restrictions: Counter-intuitively, their generalization bounds were still variant for the function-preserving parameter scaling transformation or limited only to an impractical network structure. As a more fundamental solution, we propose new prior and posterior distributions invariant to scaling transformations by \textit{decomposing} the scale and connectivity of parameters, thereby allowing the resulting generalization bound to describe the generalizability of a broad class of networks with the more practical class of transformations such as weight decay with batch normalization. We also show that the above issue adversely affects the uncertainty calibration of Laplace approximation and propose a solution using our invariant posterior. We empirically demonstrate our posterior provides effective flatness and calibration measures with low complexity in such a practical parameter transformation case, supporting its practical effectiveness in line with our rationale
    • …
    corecore